# Human preference optimization
Btlm 3b 8k Chat
Apache-2.0
BTLM-3B-8k-chat is a conversational version developed based on BTLM-3B-8K-base, optimized using DPO method, specifically designed for dialogue scenarios aligned with human preferences.
Large Language Model
Transformers English

B
cerebras
138
13
7B DPO Alpha
A 7B-parameter causal language model trained on multi-source datasets, optimized with DPO, supporting Chinese and English text generation tasks
Large Language Model
Transformers Supports Multiple Languages

7
CausalLM
131
54
Featured Recommended AI Models